Speaking faces for face-voice spe

نویسنده

  • Girija Chetty
چکیده

In this paper, we describe an approach for an animated speaking face synthesis and its application in modeling impostor/replay attack scenarios for face-voice based speaker verification systems. The speaking face reported here learns the spatiotemporal relationship between speech acoustics and MPEG4 compliant facial animation points. The influence of articulatory, perceptual, and prosodic acoustic features along with auditory context on prediction accuracy was examined. The results are indicative of vulnerability of audiovisual identity verification systems to impostor/replay attacks using synthetic faces. The level of vulnerability depends on several factors, such as the type of audiovisual features, the fusion techniques used for the audio and video features and their relative robustness. Also, the success of the synthetic impostor depends on the type of coarticulation models and acoustic features used for the audiovisual mapping of speaking face synthesis.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

The development of infant learning about specific face-voice relations.

This study examined the development of infants' ability to perceive, learn, and remember the unique face-voice relations of unfamiliar adults. Infants of 2, 4, and 6 months were habituated to the faces and voices of 2 same-gender adults speaking and then received test trials where the faces and voices were synchronized yet mismatched. Results indicated that 4- and 6-month-olds, but not 2-month-...

متن کامل

Intermodal perception of adult and child faces and voices by infants.

This research investigated the ability of 4- and 7-month-old infants to match unfamiliar, dynamic faces and voices on the basis of age or maturity. In Experiment 1, infants received videotaped trials of an adult and a child of the same gender, side by side, speaking a nursery rhyme in synchrony with one another. The voice to one and then the other face was played in synchrony with the movements...

متن کامل

Static vs. Dynamic Faces as Retrieval Cues in Recognition of Spoken Words

Three experiments examined the integration of auditory and visual information in memory for spoken words. Across experiments, recognition of isolated words was tested in the context of studied or non-studied faces and voices. The degree to which faces were informative about the studied speech event was manipulated between experiments. In Experiment 1, faces were static pictures experimentally p...

متن کامل

Electrophysiological correlates of recollecting faces of known and unknown individuals.

We recorded brain potentials from healthy human subjects during a recognition test in order to monitor neural processing associated with face recollection. Subjects first attempted to memorize 40 faces; half were accompanied by a voice simulating that person speaking (e.g., "I'm Jimmy and I was a roadie for the Grateful Dead") and half were presented in silence. In the test phase, subjects atte...

متن کامل

Seeing Voices and Hearing Faces: Cross-modal biometric matching

We introduce a seemingly impossible task: given only an audio clip of someone speaking, decide which of two face images is the speaker. In this paper we study this, and a number of related cross-modal tasks, aimed at answering the question: how much can we infer from the voice about the face and vice versa? We study this task “in the wild”, employing the datasets that are now publicly available...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2006